2 research outputs found
Selective Adaptation in Speech: Measuring the Effects of Visual and Lexical Contexts
Published Aug 1, 2021Speech selective adaptation is a phenomenon in which repeated presentation of a speech stimulus alters
subsequent phonetic categorization. Prior work has reported that lexical, but not multisensory, context
influences selective adaptation. This dissociation suggests that lexical and multisensory contexts influence
speech perception through separate and independent processes (see Samuel & Lieblich, 2014).
However, this dissociation is based on results reported by different studies using different stimuli. This
leaves open the possibility that the divergent effects of multisensory and lexical contexts on selective
adaptation may be the result of idiosyncratic differences in the stimuli rather than separate perceptual
processes. The present investigation used a single stimulus set to compare the selective adaptation produced
by lexical and multisensory contexts. In contrast to the apparent dissociation in the literature, we
find that multisensory information can in fact support selective adaptation.Support for this project was provided by NSF Grant 1632530 to
Lawrence D. Rosenblum as well as the Spanish Ministry of Science and
Innovation, Grant PSI2017-82563-P, awarded to Arthur G. Samuel and
was partially supported by the Basque Government through the BERC
2018-2021 program, and by the Spanish State Research Agency through
BCBL Severo Ochoa excellence accreditation SEV-2015-0490 awarded to
Arthur G. Samuel
The Benefit of Bimodal Training in Voice Learning
It is known that talkers can be recognized by listening to their specific vocal qualities—breathiness and fundamental frequencies. However, talker identification can also occur by focusing on the talkers’ unique articulatory style, which is known to be available auditorily and visually and can be shared across modalities. Evidence shows that voices heard while seeing talkers’ faces are later recognized better on their own compared to the voices heard alone. The present study investigated whether the facilitation of voice learning through facial cues relies on talker-specific articulatory or nonarticulatory facial information. Participants were initially trained to learn the voices of ten talkers presented either on their own or together with (a) an articulating face, (b) a static face, or (c) an isolated articulating mouth. Participants were then tested on recognizing the voices on their own regardless of their training modality. Consistent with previous research, voices learned with articulating faces were recognized better on their own compared to voices learned alone. However, isolated articulating mouths did not provide an advantage in learning the voices. The results demonstrated that learning voices while seeing faces resulted in better voice learning compared to the voices learned alone